This is an interactive notebook. You can run it locally or use the links below:
It is often hard to automatically evaluate a generated LLM response so, depending on your risk tolerance, you can gather direct user feedback to find areas to improve. In this tutorial, we’ll use a custom chatbot as an example app from which to collect user feedback. We’ll use Streamlit to build the interface and we’ll capture the LLM interactions and feedback in Weave.
Setup
chatbot.py
with the following contents:
streamlit run chatbot.py
.
Now, you can interact with this application and click the feedback buttons after each response.
Visit the Weave UI to see the attached feedback.
Explanation
If we consider our decorated prediction function as:call
object, which is obtained by using the .call()
method instead of calling the function as normal:
result
above.